FNA++: Fast Network Adaptation via Parameter Remapping and Architecture Search
نویسندگان
چکیده
Deep neural networks achieve remarkable performance in many computer vision tasks. Most state-of-the-art (SOTA) semantic segmentation and object detection approaches reuse network architectures designed for image classification as the backbone, commonly pre-trained on ImageNet. However, gains can be achieved by designing specifically segmentation, shown recent architecture search (NAS) research segmentation. One major challenge though is that ImageNet pre-training of space representation (a.k.a. super network) or searched incurs huge computational cost. In this paper, we propose a Fast Network Adaptation (FNA++) method, which adapt both parameters seed (e.g., an to become with different depths, widths, kernel sizes via parameter remapping technique, making it possible use NAS tasks lot more efficiently. our experiments, apply FNA++ MobileNetV2 obtain new detection, human pose estimation clearly outperform existing manually NAS. We also implement ResNets networks, demonstrates great generalization ability. The total computation cost significantly less than SOTA approaches: 1737× DPC, 6.8× Auto-DeepLab, 8.0× DetNAS. A series ablation studies are performed demonstrate effectiveness, detailed analysis provided insights into working mechanism. Codes available at https://github.com/JaminFong/FNA.
منابع مشابه
Efficient Neural Architecture Search via Parameter Sharing
We propose Efficient Neural Architecture Search (ENAS), a fast and inexpensive approach for automatic model design. In ENAS, a controller discovers neural network architectures by searching for an optimal subgraph within a large computational graph. The controller is trained with policy gradient to select a subgraph that maximizes the expected reward on a validation set. Meanwhile the model cor...
متن کاملArchitecture-Dependent Loop Scheduling via Communication-Sensitive Remapping
In this paper, we propose a novel efficient technique called cyclo-compaction scheduling, taking into account the data transmission delays and loop carried dependency associated with specific target architectures. This technique uses the retiming technique (loop pipelining), implicitly applied, and a task remapping to appropriate processors in order to compact the schedule length and improve th...
متن کاملGradient-free Policy Architecture Search and Adaptation
We develop a method for policy architecture search and adaptation via gradient-free optimization which can learn to perform autonomous driving tasks. By learning from both demonstration and environmental reward we develop a model that can learn with relatively few early catastrophic failures. We first learn an architecture of appropriate complexity to perceive aspects of world state relevant to...
متن کاملDifferentiable Neural Network Architecture Search
The successes of deep learning in recent years has been fueled by the development of innovative new neural network architectures. However, the design of a neural network architecture remains a difficult problem, requiring significant human expertise as well as computational resources. In this paper, we propose a method for transforming a discrete neural network architecture space into a continu...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Pattern Analysis and Machine Intelligence
سال: 2021
ISSN: ['1939-3539', '2160-9292', '0162-8828']
DOI: https://doi.org/10.1109/tpami.2020.3044416